Visual Laughter Synthesis: Initial Approaches
نویسندگان
چکیده
Visual laughter synthesis is a challenging task that was only rarely explored and empirical investigations are scarce. For the purpose of building a virtual agent able to laugh naturally we exploit different animation techniques such as a procedural animation or based on motion capture and we apply them to visual laughter synthesis. At the moment we focus on three approaches: procedural animation based on manual annotation of facial behavior; motion capture driven animation and animation generated from automatic facial movements detection. For the purpose of this study we use the Greta agent (Niewiadomski et al., 2011) that can be driven by both high-level anatomically inspired facial behavior description based on the Facial Action Coding System (FACS; Ekman, et al., 1978) or low-level facial animation parameterization (FAPs) that is a part of MPEG-4 standard (Ostermann, 2002) for facial animation. We also use two video corpora: AVLC database (Urbain et al., 2011) containing mocap, video and audio data of 24 subjects showing spontaneous amusement laugh responses and Queen's University Belfast's dataset of same sex interaction dyads during the watching of funny stimuli. We present all approaches in detail. Manual annotation of action units. FACS is a comprehensive anatomically based system for measuring all visually discernible facial movement. It describes all distinguishable facial activity on the basis of 44 unique Action Units (AUs), as well as several categories for head and eye positions/movements and miscellaneous actions. Using FACS and viewing digital-recorded facial behavior at frame rate and in slow motion, certified FACS coders are able to distinguish and code all visually discernible facial expressions. Utilizing this technique, a selection of twenty pre-recorded, laboratory stimulated, laughter events were coded. These codes were then used to model the facial behavior on the Greta agent which is able to display any configuration of AUs. For 3 virtual characters single AUs were defined and validated by certified FACS coders under the constraints of the technology. A Behavior Markup Language (BML) implemented in Greta permits the control of each AU of the agent (its duration and intensity) independently. The animation of any AU is linearly interpolated according to Attack-Decay-Sustain-Release model (Ekman, et al., 1978). Next, the symbolic intensity values are converted to low-level facial animation parameters (FAP) which, finally, are used to deform a mesh of the virtual model. We also developed a tool that automatically converts manual annotation files created with Noldus Observer XT, a commercial tool …
منابع مشابه
Audio-visual Laughter Synthesis System
In this paper we propose an overview of a project aiming at building an audio-visual laughter synthesis system. The same approach is followed for acoustic and visual synthesis. First a database has been built to have synchronous audio and 3D visual landmarks tracking data. Then this data has been used to build HMM models of acoustic laughter and visual laughter separately. Visual laughter model...
متن کاملImitating Conversational Laughter with an Articulatory Speech Synthesizer
In this study we present initial efforts to model laughter with an articulatory speech synthesizer. We aimed at imitating a real laugh taken from a spontaneous speech database and created several synthetic versions of it using articulatory synthesis and diphone synthesis. In modeling laughter with articulatory synthesis, we also approximated features like breathing noises that do not normally o...
متن کاملThe AV-LASYN Database : A synchronous corpus of audio and 3D facial marker data for audio-visual laughter synthesis
A synchronous database of acoustic and 3D facial marker data was built for audio-visual laughter synthesis. Since the aim is to use this database for HMM-based modeling and synthesis, the amount of collected data from one given subject had to be maximized. The corpus contains 251 utterances of laughter from one male participant. Laughter was elicited with the help of humorous videos. The result...
متن کاملA Phonetic Analysis of Natural Laughter, for Use in Automatic Laughter Processing Systems
In this paper, we present the detailed phonetic annotation of the publicly available AVLaughterCycle database, which can readily be used for automatic laughter processing (analysis, classification, browsing, synthesis, etc.). The phonetic annotation is used here to analyze the database, as a first step. Unsurprisingly, we find that h-like phones and central vowels are the most frequent sounds i...
متن کاملHumanoid robots laughing in response to a joke: Results of a video-based online survey
In this paper we present first results of two online surveys designed to investigate, which kind of recorded, human laughter appears to be most naturally to a human observer in combination with body movements of two different humanoid robots. As humans most often laugh within a social context [1, 2], which in turn influences the style of their laughter, we told the participants of both surveys ...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2012